您现在的位置:首页 > 学术研究 > 论文发表 > A DNN inference acceleration algorithm combining model partition and task allocation in heterogeneous edge computing system
A DNN inference acceleration algorithm combining model partition and task allocation in heterogeneous edge computing system
[发布时间:2021-08-26  阅读次数: 540]

作者:Lei Shi, Zhigang Xu, Yabo Sun, Yi Shi, Yuqi Fan, Xu Ding

发表刊物:Peer-to-Peer Networking and Applications

年份:August 2021

摘要:Edge intelligence, as a new computing paradigm, aims to allocate Artificial Intelligence (AI)-based tasks partly on the edge to execute for reducing latency, consuming energy and improving privacy. As the most important technique of AI, Deep Neural Networks (DNNs) have been widely used in various fields. And for those DNN based tasks, a new computing scheme named DNN model partition can further reduce the execution time. This computing scheme partitions the DNN task into two parts, one will be executed on the end devices and the other will be executed on edge servers. However, in a complex edge computing system, it is difficult to coordinate DNN model partition and task allocation. In this work, we study this problem in the heterogeneous edge computing system. We first establish the mathematical model of adaptive DNN model partition and task offloading. The mathematical model contains a large number of binary variables, and the solution space will be too large to be solved directly in a multi-task scenario. Then we use dynamic programming and greedy strategy to reduce the solution space under the premise of a good solution, and propose our offline algorithm named GSPI. Then considering the actual situation, we subsequently proposed the online algorithm. Through our experiments and simulations, we proved that compared with end-only and server-only, our proposed GSPI algorithm can reduce the system time cost by 30% on average and the online algorithm can reduce the system time cost by 28% on average.

参考文献拷贝字段:Lei Shi, Zhigang Xu, Yabo Sun, Yi Shi, Yuqi Fan, Xu Ding.A DNN Inference Acceleration Algorithm Combining Model Partition and Task Allocation in Heterogeneous Edge Computing System[J]. Peer-to-Peer Networking and Applications,2021,14(6): 4031-4045.


相关下载:
    A DNN inference acceleration algorithm combining model partition and task allocation in heterogeneous edge computing system